Gradient-Based Neural Architecture Search: A Comprehensive Evaluation

نویسندگان

چکیده

One of the challenges in deep learning involves discovering optimal architecture for a specific task. This is effectively tackled through Neural Architecture Search (NAS). encompasses three prominent approaches—reinforcement learning, evolutionary algorithms, and gradient descent—that have demonstrated noteworthy potential identifying good candidate architectures. However, approaches based on reinforcement algorithms often necessitate extensive computational resources, requiring hundreds GPU days or more. Therefore, we confine this work to gradient-based approach due its lower resource demands. Our objective NAS method pinpointing opportunities future enhancements. To achieve this, comprehensive evaluation use four major Gradient descent-based search methods best neural image classification tasks provided. An overview these methods, i.e., DARTS, PDARTS, Fair DARTS Att-DARTS, presented. A theoretical comparison, spaces, continuous relaxation strategy bi-level optimization, deriving then The strong weak features are also listed. Experimental results comparing error rate cost analyzed. These experiments involved using bench marking datasets CIFAR-10, CIFAR-100 ImageNet. show that PDARTS better faster among examined making it potent automating Search. By conducting comparative analysis, our research provides valuable insights directions address criticism gaps literature.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A conjugate gradient based method for Decision Neural Network training

Decision Neural Network is a new approach for solving multi-objective decision-making problems based on artificial neural networks. Using inaccurate evaluation data, network training has improved and the number of educational data sets has decreased. The available training method is based on the gradient decent method (BP). One of its limitations is related to its convergence speed. Therefore,...

متن کامل

Progressive Neural Architecture Search

We propose a method for learning CNN structures that is more efficient than previous approaches: instead of using reinforcement learning (RL) or genetic algorithms (GA), we use a sequential model-based optimization (SMBO) strategy, in which we search for architectures in order of increasing complexity, while simultaneously learning a surrogate function to guide the search, similar to A* search....

متن کامل

A Comprehensive Framework for Architecture Evaluation

Experience and research have shown that the choice of appropriate architectures in the early design phases of software-intensive systems has a determinant impact on many of the crucial qualities of such systems, e.g., maintainability, efficiency, reusability. In addition, early wrong decisions during architectural design are in general difficult and costly to change during subsequent stages of ...

متن کامل

Gradient-free Policy Architecture Search and Adaptation

We develop a method for policy architecture search and adaptation via gradient-free optimization which can learn to perform autonomous driving tasks. By learning from both demonstration and environmental reward we develop a model that can learn with relatively few early catastrophic failures. We first learn an architecture of appropriate complexity to perceive aspects of world state relevant to...

متن کامل

Differentiable Neural Network Architecture Search

The successes of deep learning in recent years has been fueled by the development of innovative new neural network architectures. However, the design of a neural network architecture remains a difficult problem, requiring significant human expertise as well as computational resources. In this paper, we propose a method for transforming a discrete neural network architecture space into a continu...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Machine learning and knowledge extraction

سال: 2023

ISSN: ['2504-4990']

DOI: https://doi.org/10.3390/make5030060